评论包含有关产品特征和用户兴趣的丰富信息,因此通常用于提高建议系统性能。具体而言,先前的工作表明,共同学习进行审查生成可以改善评级预测性能。同时,这些模型制作的评论是推荐说明,为用户提供了有关预测评分的见解。但是,尽管现有模型可能会产生流利的人类样评论,但尚不清楚评论在多大程度上完全揭示了共同预测的评级背后的理由。在这项工作中,我们执行一系列评估,以探究最先进的模型及其审查生成部分。我们表明,生成的解释是脆弱的,需要进一步评估,然后才能作为估计评级的字面原理。
translated by 谷歌翻译
最近的模型可以产生流利和语法合成评论,同时准确预测用户评分。生成的评论表达了用户对相关产品的估计意见,通常被视为自然语言“理由”,共同预测的评级。但是,先前的研究发现,现有模型通常会产生重复性,普遍适用和通用的解释,从而导致非信息原理。此外,我们的分析表明,以前的模型生成的内容通常包含事实幻觉。这些问题要求采用新颖的解决方案,这些解决方案可以产生信息丰富的和事实扎根的解释。受到最新使用检索内容的启发,除了生成的参数知识外,我们建议用个性化的检索器增强发电机,在该发现者的启发下,猎犬的输出是增强发电机的外部知识。关于Yelp,TripAdvisor和Amazon Movie评论数据集的实验表明,我们的模型可以产生解释,即更可靠地需要进行现有评论,更多样化,并且由人类评估人员评为更有信息。
translated by 谷歌翻译
会话推荐系统提供互动,参与用户的互动方式的承诺,以查找他们喜欢的物品。我们寻求通过三维提高对话建议:1)我们的目标是模仿建议的常见人类互动模式:专家证明他们的建议,寻求者解释为什么他们不喜欢该项目,双方遍历对话框迭代对话框找到合适的物品。 2)我们利用对会话批评的想法来允许用户通过批评主观方面灵活地与自然语言理由进行互动。 3)我们将会话建议适应更广泛的域名,其中不可用的人群地面真理对话框。我们开发了一个新的两部分框架,用于培训会话推荐系统。首先,我们培训推荐制度,共同建议项目,并用主观方面证明其推理。然后,我们微调该模型通过自我监督的机器人播放来合并迭代用户反馈。三个真实数据集的实验表明,与最先进的方法相比,我们的系统可以应用于各种域的不同推荐模型,以实现对话建议的卓越性能。我们还评估了我们对人类用户的模型,显示在我们的框架下培训的系统提供更有用,有用,有用,并且在热情和冷启动设置中提供的知识推荐。
translated by 谷歌翻译
提取理由(即输入特征的子集)和自然语言解释(NLEL)是机器学习模型的两个主要类型的解释。虽然NLES可以比提取理由更全面,但是已经显示出机器生成的NLES在致致通知知识方面缩短。在本文中,我们表明,致辞知识可以充当开采理性和NLE之间的桥梁,更好地渲染两种类型的解释。我们介绍了一个名为rexc的自律化框架,(1)提取理由作为预测的最负责任的特征,(2)使用致辞资源扩展到的提取理由,(3)选择最适合的致商知识来生成NLE并提供最终预测。我们的框架在自然语言和视觉语言理解中,我们在以前的五个任务中产生了一个大型的最先进的最先进。与致致辞的自合作化也强烈提高了在以前产生解释的最佳表演模型上的提取理由和任务表演的质量。
translated by 谷歌翻译
This paper proposes a novel self-supervised based Cut-and-Paste GAN to perform foreground object segmentation and generate realistic composite images without manual annotations. We accomplish this goal by a simple yet effective self-supervised approach coupled with the U-Net based discriminator. The proposed method extends the ability of the standard discriminators to learn not only the global data representations via classification (real/fake) but also learn semantic and structural information through pseudo labels created using the self-supervised task. The proposed method empowers the generator to create meaningful masks by forcing it to learn informative per-pixel as well as global image feedback from the discriminator. Our experiments demonstrate that our proposed method significantly outperforms the state-of-the-art methods on the standard benchmark datasets.
translated by 谷歌翻译
Indian e-commerce industry has evolved over the last decade and is expected to grow over the next few years. The focus has now shifted to turnaround time (TAT) due to the emergence of many third-party logistics providers and higher customer expectations. The key consideration for delivery providers is to balance their overall operating costs while meeting the promised TAT to their customers. E-commerce delivery partners operate through a network of facilities whose strategic locations help to run the operations efficiently. In this work, we identify the locations of hubs throughout the country and their corresponding mapping with the distribution centers. The objective is to minimize the total network costs with TAT adherence. We use Genetic Algorithm and leverage business constraints to reduce the solution search space and hence the solution time. The results indicate an improvement of 9.73% in TAT compliance compared with the current scenario.
translated by 谷歌翻译
Independence testing is a fundamental and classical statistical problem that has been extensively studied in the batch setting when one fixes the sample size before collecting data. However, practitioners often prefer procedures that adapt to the complexity of a problem at hand instead of setting sample size in advance. Ideally, such procedures should (a) allow stopping earlier on easy tasks (and later on harder tasks), hence making better use of available resources, and (b) continuously monitor the data and efficiently incorporate statistical evidence after collecting new data, while controlling the false alarm rate. It is well known that classical batch tests are not tailored for streaming data settings, since valid inference after data peeking requires correcting for multiple testing, but such corrections generally result in low power. In this paper, we design sequential kernelized independence tests (SKITs) that overcome such shortcomings based on the principle of testing by betting. We exemplify our broad framework using bets inspired by kernelized dependence measures such as the Hilbert-Schmidt independence criterion (HSIC) and the constrained-covariance criterion (COCO). Importantly, we also generalize the framework to non-i.i.d. time-varying settings, for which there exist no batch tests. We demonstrate the power of our approaches on both simulated and real data.
translated by 谷歌翻译
This paper describes a simple yet efficient repetition-based modular system for speeding up air-traffic controllers (ATCos) training. E.g., a human pilot is still required in EUROCONTROL's ESCAPE lite simulator (see https://www.eurocontrol.int/simulator/escape) during ATCo training. However, this need can be substituted by an automatic system that could act as a pilot. In this paper, we aim to develop and integrate a pseudo-pilot agent into the ATCo training pipeline by merging diverse artificial intelligence (AI) powered modules. The system understands the voice communications issued by the ATCo, and, in turn, it generates a spoken prompt that follows the pilot's phraseology to the initial communication. Our system mainly relies on open-source AI tools and air traffic control (ATC) databases, thus, proving its simplicity and ease of replicability. The overall pipeline is composed of the following: (1) a submodule that receives and pre-processes the input stream of raw audio, (2) an automatic speech recognition (ASR) system that transforms audio into a sequence of words; (3) a high-level ATC-related entity parser, which extracts relevant information from the communication, i.e., callsigns and commands, and finally, (4) a speech synthesizer submodule that generates responses based on the high-level ATC entities previously extracted. Overall, we show that this system could pave the way toward developing a real proof-of-concept pseudo-pilot system. Hence, speeding up the training of ATCos while drastically reducing its overall cost.
translated by 谷歌翻译
Knowledge distillation is often used to transfer knowledge from a strong teacher model to a relatively weak student model. Traditional knowledge distillation methods include response-based methods and feature-based methods. Response-based methods are used the most widely but suffer from lower upper limit of model performance, while feature-based methods have constraints on the vocabularies and tokenizers. In this paper, we propose a tokenizer-free method liberal feature-based distillation (LEAD). LEAD aligns the distribution between teacher model and student model, which is effective, extendable, portable and has no requirements on vocabularies, tokenizer, or model architecture. Extensive experiments show the effectiveness of LEAD on several widely-used benchmarks, including MS MARCO Passage, TREC Passage 19, TREC Passage 20, MS MARCO Document, TREC Document 19 and TREC Document 20.
translated by 谷歌翻译
This paper presents E5, a family of state-of-the-art text embeddings that transfer well to a wide range of tasks. The model is trained in a contrastive manner with weak supervision signals from our curated large-scale text pair dataset (called CCPairs). E5 can be readily used as a general-purpose embedding model for any tasks requiring a single-vector representation of texts such as retrieval, clustering, and classification, achieving strong performance in both zero-shot and fine-tuned settings. We conduct extensive evaluations on 56 datasets from the BEIR and MTEB benchmarks. For zero-shot settings, E5 is the first model that outperforms the strong BM25 baseline on the BEIR retrieval benchmark without using any labeled data. When fine-tuned, E5 obtains the best results on the MTEB benchmark, beating existing embedding models with 40x more parameters.
translated by 谷歌翻译